Type :
- AI News
- AI Tools
- AI Cases
- AI Tutorial
2023-12-12 10:15:16.AIbase.4.1k
Research on the Risks of Training Data Leakage in ChatGPT
The research found that large language models like ChatGPT retain a considerable amount of original training data during their training process. Attackers can extract large volumes of training data from the model using specific attack methods, threatening the privacy of data owners. Researchers utilized a concept of 'extractable memory' attack techniques and discovered that existing extraction attacks can recover a significant amount of training data from memory even without using the actual training data as prompts. ⚠️ Researchers recommend that data security protective measures should be implemented while developing and using large language models.
2023-12-04 16:22:05.AIbase.3.8k
AI Researchers Successfully Reveal ChatGPT's Training Data Sources Through 'Poetry Attack'
Teams from Google DeepMind, the University of Washington, and others have successfully unveiled the sources of ChatGPT's training data through a 'poetry attack'. This attack method involves repeating specific vocabulary, which led ChatGPT to disclose personal information, including phone numbers and email addresses. The vulnerability was reported to OpenAI on August 30, and no further comments have been made. The research calls for environmentally-friendly AI usage, pointing out the significant energy burden that text and image generation imposes on the environment.